Update cudnn semantic version in README to avoid performance issue #1239
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
I suggest changing the version spec of
nvidia-cublas-cu12
andnvidia-cudnn-cu12
. Using existing one:pip install nvidia-cublas-cu12 nvidia-cudnn-cu12==9.*
on 2025-02-07 gives
which results in cuda running 30x slower than cpu on my Ubuntu 22.04, with a 6 seconds wav audio.
After I downgrade them to:
which is what
torch==2.6.0
currently depends on, everything works fine. Then, I go through a thorough grid search over recent cublas and cudnn versions, and find that all cudnn after 9.6 (inclusive) exhibit the aforementioned performance issue.I didn't study the lower bound, though. Hence, I make it compatible with current semantic version, i.e.
>=9
).That's why I make this change.
Thank you so much!